Goto

Collaborating Authors

 social platform


Agent Guide: A Simple Agent Behavioral Watermarking Framework

Huang, Kaibo, Zhang, Zipei, Yang, Zhongliang, Zhou, Linna

arXiv.org Artificial Intelligence

The increasing deployment of intelligent agents in digital ecosystems, such as social media platforms, has raised significant concerns about traceability and accountability, particularly in cybersecurity and digital content protection. Traditional large language model (LLM) watermarking techniques, which rely on token-level manipulations, are ill-suited for agents due to the challenges of behavior tokenization and information loss during behavior-to-action translation. To address these issues, we propose Agent Guide, a novel behavioral watermarking framework that embeds watermarks by guiding the agent's high-level decisions (behavior) through probability biases, while preserving the naturalness of specific executions (action). Our approach decouples agent behavior into two levels, behavior (e.g., choosing to bookmark) and action (e.g., bookmarking with specific tags), and applies watermark-guided biases to the behavior probability distribution. We employ a z-statistic-based statistical analysis to detect the watermark, ensuring reliable extraction over multiple rounds. Experiments in a social media scenario with diverse agent profiles demonstrate that Agent Guide achieves effective watermark detection with a low false positive rate. Our framework provides a practical and robust solution for agent watermarking, with applications in identifying malicious agents and protecting proprietary agent systems.


MPPFND: A Dataset and Analysis of Detecting Fake News with Multi-Platform Propagation

Zhao, Congyuan, Wei, Lingwei, Qin, Ziming, Zhou, Wei, Song, Yunya, Hu, Songlin

arXiv.org Artificial Intelligence

Most existing detection algorithms focus on analyzing news content and social context to detect fake news. However, these approaches typically detect fake news based on specific platforms, ignoring differences in propagation characteristics across platforms. In this paper, we introduce the MPPFND dataset, which captures propagation structures across multiple platforms. We also describe the commenting and propagation characteristics of different platforms to show that their social contexts have distinct features. We propose a multi-platform fake news detection model (APSL) that uses graph neural networks to extract social context features from various platforms. Experiments show that accounting for cross-platform propagation differences improves fake news detection performance.


Will A.I. Save the News?

The New Yorker

I am a forty-five-year-old journalist who, for many years, didn't read the news. In high school, I knew about events like the O. J. Simpson trial and the Oklahoma City bombing, but not much else. In college, I was friends with geeky economics majors who read The Economist, but I'm pretty sure I never actually turned on CNN or bought a paper at the newsstand. I read novels, and magazines like Wired and Spin. If I went online, it wasn't to check the front page of the Times but to browse record reviews from College Music Journal. Somehow, during this time, I thought of myself as well informed.


Simulation of Language Evolution under Regulated Social Media Platforms: A Synergistic Approach of Large Language Models and Genetic Algorithms

Cai, Jinyu, Ishimizu, Yusei, Zhang, Mingyue, Li, Munan, Li, Jialong, Tei, Kenji

arXiv.org Artificial Intelligence

Social media platforms frequently impose restrictive policies to moderate user content, prompting the emergence of creative evasion language strategies. This paper presents a multi-agent framework based on Large Language Models (LLMs) to simulate the iterative evolution of language strategies under regulatory constraints. In this framework, participant agents, as social media users, continuously evolve their language expression, while supervisory agents emulate platform-level regulation by assessing policy violations. To achieve a more faithful simulation, we employ a dual design of language strategies (constraint and expression) to differentiate conflicting goals and utilize an LLM-driven GA (Genetic Algorithm) for the selection, mutation, and crossover of language strategies. The framework is evaluated using two distinct scenarios: an abstract password game and a realistic simulated illegal pet trade scenario. Experimental results demonstrate that as the number of dialogue rounds increases, both the number of uninterrupted dialogue turns and the accuracy of information transmission improve significantly. Furthermore, a user study with 40 participants validates the real-world relevance of the generated dialogues and strategies. Moreover, ablation studies validate the importance of the GA, emphasizing its contribution to long-term adaptability and improved overall results.


Characterizing User Archetypes and Discussions on Scored.co

Failla, Andrea, Citraro, Salvatore, Rossetti, Giulio, Cauteruccio, Francesco

arXiv.org Artificial Intelligence

In recent years, the proliferation of social platforms has drastically transformed the way individuals interact, organize, and share information. In this scenario, we experience an unprecedented increase in the scale and complexity of interactions and, at the same time, little to no research about some fringe social platforms. In this paper, we present a multi-dimensional framework for characterizing nodes and hyperedges in social hypernetworks, with a focus on the understudied alt-right platform Scored.co. Our approach integrates the possibility of studying higher-order interactions, thanks to the hypernetwork representation, and various node features such as user activity, sentiment, and toxicity, with the aim to define distinct user archetypes and understand their roles within the network. Utilizing a comprehensive dataset from Scored.co, we analyze the dynamics of these archetypes over time and explore their interactions and influence within the community. The framework's versatility allows for detailed analysis of both individual user behaviors and broader social structures. Our findings highlight the importance of higher-order interactions in understanding social dynamics, offering new insights into the roles and behaviors that emerge in complex online environments.


Exploring the Use of Abusive Generative AI Models on Civitai

Wei, Yiluo, Zhu, Yiming, Hui, Pan, Tyson, Gareth

arXiv.org Artificial Intelligence

The rise of generative AI is transforming the landscape of digital imagery, and exerting a significant influence on online creative communities. This has led to the emergence of AI-Generated Content (AIGC) social platforms, such as Civitai. These distinctive social platforms allow users to build and share their own generative AI models, thereby enhancing the potential for more diverse artistic expression. Designed in the vein of social networks, they also provide artists with the means to showcase their creations (generated from the models), engage in discussions, and obtain feedback, thus nurturing a sense of community. Yet, this openness also raises concerns about the abuse of such platforms, e.g., using models to disseminate deceptive deepfakes or infringe upon copyrights. To explore this, we conduct the first comprehensive empirical study of an AIGC social platform, focusing on its use for generating abusive content. As an exemplar, we construct a comprehensive dataset covering Civitai, the largest available AIGC social platform. Based on this dataset of 87K models and 2M images, we explore the characteristics of content and discuss strategies for moderation to better govern these platforms.


Explainability and Hate Speech: Structured Explanations Make Social Media Moderators Faster

Calabrese, Agostina, Neves, Leonardo, Shah, Neil, Bos, Maarten W., Ross, Björn, Lapata, Mirella, Barbieri, Francesco

arXiv.org Artificial Intelligence

Content moderators play a key role in keeping the conversation on social media healthy. While the high volume of content they need to judge represents a bottleneck to the moderation pipeline, no studies have explored how models could support them to make faster decisions. There is, by now, a vast body of research into detecting hate speech, sometimes explicitly motivated by a desire to help improve content moderation, but published research using real content moderators is scarce. In this work we investigate the effect of explanations on the speed of real-world moderators. Our experiments show that while generic explanations do not affect their speed and are often ignored, structured explanations lower moderators' decision making time by 7.4%.


AI meme wars hit India election campaign, testing social platforms

Al Jazeera

Bengaluru, India – On February 20, India's chief opposition party, the Indian National Congress (INC), uploaded a video parodying Prime Minister Narendra Modi on Instagram that has amassed over 1.5 million views. It is a short clip from a new Hindi music album named "Chor" (thief), where Modi's digital likeness is grafted onto the lead singer. The song's lyrics were humorously reworked to describe a thief's – in this case, a business tycoon's – attempt to steal, and Modi handing over coal mines, ports, power lines and ultimately, the country. The video isn't hyperrealistic, but a pithy AI meme that uses Modi's voice and face clones, to drive home the nagging criticism of his close ties to Indian business moguls. That same day, the official Bharatiya Janata Party (BJP) handle on Instagram, with over seven million followers, uploaded its own video.


In Defense of Humanity

The Atlantic - Technology

On July 13, 1833, during a visit to the Cabinet of Natural History at the Jardin des Plantes, in Paris, Ralph Waldo Emerson had an epiphany. Peering at the museum's specimens--butterflies, hunks of amber and marble, carved seashells--he felt overwhelmed by the interconnectedness of nature, and humankind's place within it. Check out more from this issue and find your next story to read. The experience inspired him to write "The Uses of Natural History," and to articulate a philosophy that put naturalism at the center of intellectual life in a technologically chaotic age--guiding him, along with the collective of writers and radical thinkers known as transcendentalists, to a new spiritual belief system. Through empirical observation of the natural world, Emerson believed, anyone could become "a definer and map-maker of the latitudes and longitudes of our condition"--finding agency, individuality, and wonder in a mechanized age. America was crackling with invention in those years, and everything seemed to be speeding up as a result.


Predicting Hateful Discussions on Reddit using Graph Transformer Networks and Communal Context

Hebert, Liam, Golab, Lukasz, Cohen, Robin

arXiv.org Artificial Intelligence

We propose a system to predict harmful discussions on social media platforms. Our solution uses contextual deep language models and proposes the novel idea of integrating state-of-the-art Graph Transformer Networks to analyze all conversations that follow an initial post. This framework also supports adapting to future comments as the conversation unfolds. In addition, we study whether a community-specific analysis of hate speech leads to more effective detection of hateful discussions. We evaluate our approach on 333,487 Reddit discussions from various communities. We find that community-specific modeling improves performance two-fold and that models which capture wider-discussion context improve accuracy by 28\% (35\% for the most hateful content) compared to limited context models.